A Provable Framework of Learning Graph Embeddings via Summarization
نویسندگان
چکیده
Given a large graph, can we learn its node embeddings from smaller summary graph? What is the relationship between learned original graphs and their graphs? Graph representation learning plays an important role in many graph mining applications, but em-beddings of large-scale remains challenge. Recent works try to alleviate it via summarization, which typ-ically includes three steps: reducing size by combining nodes edges into supernodes superedges,learning supernode embedding on then restoring nodes. How-ever, justification behind those steps still unknown. In this work, propose GELSUMM, well-formulated framework based sum-marization, show theoretical ground learn-ing restoration with well-known approaches closed form.Through extensive experiments real-world datasets, demonstrate that our methods matching or better performance downstream tasks.This work provides analysis for summarization helps explain under-stand mechanism existing works.
منابع مشابه
Graph Hybrid Summarization
One solution to process and analysis of massive graphs is summarization. Generating a high quality summary is the main challenge of graph summarization. In the aims of generating a summary with a better quality for a given attributed graph, both structural and attribute similarities must be considered. There are two measures named density and entropy to evaluate the quality of structural and at...
متن کاملLearning Role-based Graph Embeddings
Random walks are at the heart of many existing network embedding methods. However, such algorithms have many limitations that arise from the use of random walks, e.g., the features resulting from these methods are unable to transfer to new nodes and graphs as they are tied to vertex identity. In this work, we introduce the Role2Vec framework which uses the flexible notion of attributed random w...
متن کاملLimitations of Learning Via Embeddings
This paper considers the embeddability of general concept classes in Euclidean half spaces. By embedding in half spaces we refer to a mapping from some concept class to half spaces so that the labeling given to points in the instance space is retained. The existence of an embedding for some class may be used to learn it using an algorithm for the class it is embedded into. The Support Vector Ma...
متن کاملProvable Dictionary Learning via Column Signatures
In dictionary learning, also known as sparse coding, we are given samples of the form y = Ax where x ∈ R is an unknown random sparse vector and A is an unknown dictionary matrix in Rn×m (usually m > n, which is the overcomplete case). The goal is to learn A and x. This problem has been studied in neuroscience, machine learning, vision, and image processing. In practice it is solved by heuristic...
متن کاملRevealing Biological Modules via Graph Summarization
The division of a protein interaction network into biologically meaningful modules can aid with automated detection of protein complexes and prediction of biological processes and can uncover the global organization of the cell. We propose the use of a graph summarization (GS) technique, based on graph compression, to cluster protein interaction graphs into biologically relevant modules. The me...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2023
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v37i4.25621